Compressed Least-Squares Regression

نویسندگان

  • Odalric-Ambrym Maillard
  • Rémi Munos
چکیده

We consider the problem of learning, from K data, a regression function in a linear space of high dimensionN using projections onto a random subspace of lower dimension M . From any algorithm minimizing the (possibly penalized) empirical risk, we provide bounds on the excess risk of the estimate computed in the projected subspace (compressed domain) in terms of the excess risk of the estimate built in the high-dimensional space (initial domain). We show that solving the problem in the compressed domain instead of the initial domain reduces the estimation error at the price of an increased (but controlled) approximation error. We apply the analysis to Least-Squares (LS) regression and discuss the excess risk and numerical complexity of the resulting “Compressed Least Squares Regression” (CLSR) in terms of N , K, and M . When we choose M = O( √ K), we show that CLSR has an estimation error of order O(logK/ √ K).

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Compressed Least Squares Regression revisited

We revisit compressed least squares (CLS) regression as originally analyzed in Maillard and Munos (2009) and later on in Kaban (2014) with some refinements. Given a set of high-dimensional inputs, CLS applies a random projection and then performs least squares regression based on the projected inputs of lower dimension. This approach can be beneficial with regard to both computation (yielding a...

متن کامل

ar X iv : 0 70 6 . 05 34 v 1 [ st at . M L ] 4 J un 2 00 7 Compressed Regression

Recent research has studied the role of sparsity in high dimensional regression and signal reconstruction, establishing theoretical limits for recovering sparse models from sparse data. This line of work shows that l1-regularized least squares regression can accurately estimate a sparse linear model from n noisy examples in p dimensions, even if p is much larger than n. In this paper we study a...

متن کامل

Annotated Bibliography High-dimensional Statistical Inference

Recent research has studied the role of sparsity in high dimensional regression and signal reconstruction, establishing theoretical limits for recovering sparse models from sparse data. This line of work shows that l1-regularized least squares regression can accurately estimate a sparse linear model from n noisy examples in p dimensions, even if p is much larger than n. In this paper we study a...

متن کامل

An Enhanced Dimension Reduction Approach for Microarray Gene Expression Data

Methods: Here, we introduce an enhancement concept that joins two dimension reduction methods of Partial Least Squares (PLS) and Minimum Average Variance Estimation (MAVE), which is called as an enhanced dimension reduction method. The PLS method generates the new transformed genes that include compressed information. Then, the MAVE method clusters samples in the same class and separate samples...

متن کامل

Title Placeholder

This paper compares and discusses four techniques for model order reduction based on compressed sensing (CS), less relevant basis removal (LRBR), principal component analysis (PCA) and partial least squares (PLS). CS and PCA have already been used for reducing the order of power amplifier (PA) behavioral models for digital predistortion (DPD) purposes. While PLS, despite being popular in some s...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2009